Harvesting question-answer (QA) pairs from customer service chatlog in the wild is an efficient way to enrich the knowledge base for customer service chatbots in the cold start or continuous integration scenarios. Prior work attempts to obtain 1-to-1 QA pairs from growing customer service chatlog, which fails to integrate the incomplete utterances from the dialog context for composite QA retrieval. In this paper, we propose N-to-N QA extraction task in which the derived questions and corresponding answers might be separated across different utterances. We introduce a suite of generative/discriminative tagging based methods with end-to-end and two-stage variants that perform well on 5 customer service datasets and for the first time setup a benchmark for N-to-N DialogQAE with utterance and session level evaluation metrics. With a deep dive into extracted QA pairs, we find that the relations between and inside the QA pairs can be indicators to analyze the dialogue structure, e.g. information seeking, clarification, barge-in and elaboration. We also show that the proposed models can adapt to different domains and languages, and reduce the labor cost of knowledge accumulation in the real-world product dialogue platform.
translated by 谷歌翻译
The security of artificial intelligence (AI) is an important research area towards safe, reliable, and trustworthy AI systems. To accelerate the research on AI security, the Artificial Intelligence Security Competition (AISC) was organized by the Zhongguancun Laboratory, China Industrial Control Systems Cyber Emergency Response Team, Institute for Artificial Intelligence, Tsinghua University, and RealAI as part of the Zhongguancun International Frontier Technology Innovation Competition (https://www.zgc-aisc.com/en). The competition consists of three tracks, including Deepfake Security Competition, Autonomous Driving Security Competition, and Face Recognition Security Competition. This report will introduce the competition rules of these three tracks and the solutions of top-ranking teams in each track.
translated by 谷歌翻译
Time series forecasting is a long-standing challenge due to the real-world information is in various scenario (e.g., energy, weather, traffic, economics, earthquake warning). However some mainstream forecasting model forecasting result is derailed dramatically from ground truth. We believe it's the reason that model's lacking ability of capturing frequency information which richly contains in real world datasets. At present, the mainstream frequency information extraction methods are Fourier transform(FT) based. However, use of FT is problematic due to Gibbs phenomenon. If the values on both sides of sequences differ significantly, oscillatory approximations are observed around both sides and high frequency noise will be introduced. Therefore We propose a novel frequency enhanced channel attention that adaptively modelling frequency interdependencies between channels based on Discrete Cosine Transform which would intrinsically avoid high frequency noise caused by problematic periodity during Fourier Transform, which is defined as Gibbs Phenomenon. We show that this network generalize extremely effectively across six real-world datasets and achieve state-of-the-art performance, we further demonstrate that frequency enhanced channel attention mechanism module can be flexibly applied to different networks. This module can improve the prediction ability of existing mainstream networks, which reduces 35.99% MSE on LSTM, 10.01% on Reformer, 8.71% on Informer, 8.29% on Autoformer, 8.06% on Transformer, etc., at a slight computational cost ,with just a few line of code. Our codes and data are available at https://github.com/Zero-coder/FECAM.
translated by 谷歌翻译
Real-time semantic segmentation has played an important role in intelligent vehicle scenarios. Recently, numerous networks have incorporated information from multi-size receptive fields to facilitate feature extraction in real-time semantic segmentation tasks. However, these methods preferentially adopt massive receptive fields to elicit more contextual information, which may result in inefficient feature extraction. We believe that the elaborated receptive fields are crucial, considering the demand for efficient feature extraction in real-time tasks. Therefore, we propose an effective and efficient architecture termed Dilation-wise Residual segmentation (DWRSeg), which possesses different sets of receptive field sizes within different stages. The architecture involves (i) a Dilation-wise Residual (DWR) module for extracting features based on different scales of receptive fields in the high level of the network; (ii) a Simple Inverted Residual (SIR) module that uses an inverted bottleneck structure to extract features from the low stage; and (iii) a simple fully convolutional network (FCN)-like decoder for aggregating multiscale feature maps to generate the prediction. Extensive experiments on the Cityscapes and CamVid datasets demonstrate the effectiveness of our method by achieving a state-of-the-art trade-off between accuracy and inference speed, in addition to being lighter weight. Without using pretraining or resorting to any training trick, we achieve 72.7% mIoU on the Cityscapes test set at a speed of 319.5 FPS on one NVIDIA GeForce GTX 1080 Ti card, which is significantly faster than existing methods. The code and trained models are publicly available.
translated by 谷歌翻译
关于对比学习的最新研究仅通过在医学图像分割的背景下利用很少的标签来实现出色的性能。现有方法主要关注实例歧视和不变映射。但是,他们面临三个常见的陷阱:(1)尾巴:医疗图像数据通常遵循隐式的长尾分配。盲目利用训练中的所有像素会导致数据失衡问题,并导致性能恶化; (2)一致性:尚不清楚分割模型是否由于不同解剖学特征之间的类内变化而学会了有意义但一致的解剖学特征; (3)多样性:整个数据集中的切片内相关性已得到明显降低的关注。这促使我们寻求一种有原则的方法来战略利用数据集本身,以发现不同解剖学观点的类似但不同的样本。在本文中,我们介绍了一种新型的半监督医学图像分割框架,称其为您自己的解剖结构(MONA),并做出了三个贡献。首先,先前的工作认为,每个像素对模型培训都同样重要。我们从经验上观察到,仅此单单就不太可能定义有意义的解剖特征,这主要是由于缺乏监督信号。我们通过使用更强大的数据增强和最近的邻居展示了学习不变的两个简单解决方案。其次,我们构建了一组目标,鼓励模型能够以无监督的方式将医学图像分解为解剖特征的集合。最后,我们在具有不同标记设置的三个基准数据集上的广泛结果验证了我们提出的MONA的有效性,该数据在不同的标签设置下实现了新的最新设置。
translated by 谷歌翻译
从RGB-D图像中对刚性对象的6D姿势估计对于机器人技术中的对象抓握和操纵至关重要。尽管RGB通道和深度(d)通道通常是互补的,分别提供了外观和几何信息,但如何完全从两个跨模式数据中完全受益仍然是非平凡的。从简单而新的观察结果来看,当对象旋转时,其语义标签是姿势不变的,而其关键点偏移方向是姿势的变体。为此,我们提出了So(3)pose,这是一个新的表示学习网络,可以探索SO(3)equivariant和So(3) - 从深度通道中进行姿势估计的特征。 SO(3) - 激素特征有助于学习更独特的表示,以分割来自RGB通道外观相似的对象。 SO(3) - 等级特征与RGB功能通信,以推导(缺失的)几何形状,以检测从深度通道的反射表面的对象的关键点。与大多数现有的姿势估计方法不同,我们的SO(3) - 不仅可以实现RGB和深度渠道之间的信息通信,而且自然会吸收SO(3) - 等级的几何学知识,从深度图像中,导致更好的外观和更好的外观和更好几何表示学习。综合实验表明,我们的方法在三个基准测试中实现了最先进的性能。
translated by 谷歌翻译
我们考虑对具有3D结构的蛋白质的代表性学习。我们基于蛋白质结构构建3D图并开发图形网络以学习其表示形式。根据我们希望捕获的细节级别,可以在不同级别计算蛋白质表示,\ emph {e.g。},氨基酸,骨干或全原子水平。重要的是,不同级别之间存在层次关系。在这项工作中,我们建议开发一个新型的层次图网络(称为pronet)来捕获关系。我们的pronet非常灵活,可用于计算不同水平粒度的蛋白质表示。我们表明,鉴于完整的基本3D图网络,我们的PRONET表示在所有级别上也已完成。为了关闭循环,我们开发了一个完整有效的3D图网络,以用作基本模型,从而使我们的pronet完成。我们对多个下游任务进行实验。结果表明,PRONET优于大多数数据集上的最新方法。此外,结果表明,不同的下游任务可能需要不同级别的表示。我们的代码可作为DIG库的一部分(\ url {https://github.com/divelab/dig})。
translated by 谷歌翻译
无监督的域适应性(UDA)方法已广泛用于提高模型在一般计算机视觉中的适应能力。但是,与自然图像不同,在组织病理学图像中不同类别的核存在巨大的语义差距。它仍未探索,我们如何构建通用的UDA模型来精确分割或分类不同数据集的核实例。在这项工作中,我们提出了一个新颖的深神经网络,即用于UDA Nuclei实例分割和分类的类别感知特征对齐和伪标记网络(CAPL-NET)。具体而言,我们首先提出一个具有动态可学习权衡权重的类别级特征对齐模块。其次,我们建议通过基于Nuclei-Level原型特征的伪标签来促进目标数据上的模型性能。关于跨域核实例分割和分类任务的综合实验表明,我们的方法优于最先进的UDA方法。
translated by 谷歌翻译
许多现实世界数据可以建模为3D图,但是完全有效地包含3D信息的学习表示形式具有挑战性。现有方法要么使用部分3D信息,要么遭受过多的计算成本。为了完全有效地合并3D信息,我们提出了一个新的消息传递方案,该方案在1跳社区内运行。我们的方法通过实现全球和本地完整性来确保有关3D图的3D信息的完整性。值得注意的是,我们提出了重要的旋转角度来实现全球完整性。此外,我们证明我们的方法比先前的方法快。我们为我们的方法提供了严格的完整性证明和时间复杂性的分析。由于分子本质上是量子系统,我们通过梳理量子启发的基础函数和提出的消息传递方案来构建\下划线{com} plete {com} plete {com} plete {com} plete {e}。实验结果证明了COMENET的能力和效率,尤其是在数量和尺寸大小的现实数据集上。我们的代码作为DIG库的一部分公开可用(\ url {https://github.com/divelab/dig})。
translated by 谷歌翻译
机器学习在解决无线干扰管理问题方面取得了成功。已经培训了不同种类的深神经网络(DNN),以完成功率控制,波束成形和准入控制等关键任务。基于DNNS的干扰管理模型有两个流行的培训范式:监督学习(即,由优化算法产生的拟合标签)和无监督的学习(即,直接优化一些系统性能测量)。虽然这两种范式都在实践中广泛应用,但由于对这些方法缺乏任何理论理解,但目前尚不清楚如何系统地理解和比较他们的性能。在这项工作中,我们开展理论研究,为这两个训练范例提供了一些深入的了解。首先,我们展示了一些令人惊讶的结果,即对于一些特殊的功率控制问题,无监督的学习可以表现比监督对手更糟糕,因为它更有可能陷入一些低质量的本地解决方案。然后,我们提供了一系列理论结果,以进一步了解两种方法的性质。一般来说,我们表明,当有高质量的标签可用时,监督学习不太可能陷入解决方案,而不是无监督的对应物。此外,我们开发了一种半监督的学习方法,可以妥善整合这两个训练范例,可以有效地利用有限数量的标签来找到高质量的解决方案。为了我们的知识,这些是第一种在基于学习的无线通信系统设计中了解不同培训方法的第一组理论结果。
translated by 谷歌翻译